我们提出了MDEAW,这是一个多模式数据库,该数据库由电动活动(EDA)和光摄影学(PPG)信号组成,在考试期间记录了巴塞罗那萨巴德尔(Eurecat Academemy)的老师教师教授的课程,以引起对学生对学生对情感反应的情感反应。课堂场景。以6种基本的情感状态来记录了10名学生的信号以及学生对每个刺激后对情感状态的自我评估。所有信号均使用便携式,可穿戴,无线,低成本和现成的设备捕获,该设备有可能在日常应用中使用情感计算方法。使用基于EDA和PPG的功能及其融合的学生识别的基线是通过remecs,fed-emecs和fed-emecs-u建立的。这些结果表明,使用低成本设备进行情感状态识别应用的前景。提出的数据库将公开可用,以使研究人员能够对这些捕获设备对情绪状态识别应用的适用性进行更透彻的评估。
translated by 谷歌翻译
This paper introduces a novel algorithm, the Perturbed Proximal Preconditioned SPIDER algorithm (3P-SPIDER), designed to solve finite sum non-convex composite optimization. It is a stochastic Variable Metric Forward-Backward algorithm, which allows approximate preconditioned forward operator and uses a variable metric proximity operator as the backward operator; it also proposes a mini-batch strategy with variance reduction to address the finite sum setting. We show that 3P-SPIDER extends some Stochastic preconditioned Gradient Descent-based algorithms and some Incremental Expectation Maximization algorithms to composite optimization and to the case the forward operator can not be computed in closed form. We also provide an explicit control of convergence in expectation of 3P-SPIDER, and study its complexity in order to satisfy the epsilon-approximate stationary condition. Our results are the first to combine the composite non-convex optimization setting, a variance reduction technique to tackle the finite sum setting by using a minibatch strategy and, to allow deterministic or random approximations of the preconditioned forward operator. Finally, through an application to inference in a logistic regression model with random effects, we numerically compare 3P-SPIDER to other stochastic forward-backward algorithms and discuss the role of some design parameters of 3P-SPIDER.
translated by 谷歌翻译
As AI systems become more capable, we would like to enlist their help to supervise other AIs. We experiment with methods for training a harmless AI assistant through self-improvement, without any human labels identifying harmful outputs. The only human oversight is provided through a list of rules or principles, and so we refer to the method as 'Constitutional AI'. The process involves both a supervised learning and a reinforcement learning phase. In the supervised phase we sample from an initial model, then generate self-critiques and revisions, and then finetune the original model on revised responses. In the RL phase, we sample from the finetuned model, use a model to evaluate which of the two samples is better, and then train a preference model from this dataset of AI preferences. We then train with RL using the preference model as the reward signal, i.e. we use 'RL from AI Feedback' (RLAIF). As a result we are able to train a harmless but non-evasive AI assistant that engages with harmful queries by explaining its objections to them. Both the SL and RL methods can leverage chain-of-thought style reasoning to improve the human-judged performance and transparency of AI decision making. These methods make it possible to control AI behavior more precisely and with far fewer human labels.
translated by 谷歌翻译
Edge computing is becoming more and more popular among researchers who seek to take advantage of the edge resources and the minimal time delays, in order to run their robotic applications more efficiently. Recently, many edge architectures have been proposed, each of them having their advantages and disadvantages, depending on each application. In this work, we present two different edge architectures for controlling the trajectory of an Unmanned Aerial Vehicle (UAV). The first architecture is based on docker containers and the second one is based on kubernetes, while the main framework for operating the robot is the Robotic Operating System (ROS). The efficiency of the overall proposed scheme is being evaluated through extended simulations for comparing the two architectures and the overall results obtained.
translated by 谷歌翻译
Large language models (LLMs) have been shown to be able to perform new tasks based on a few demonstrations or natural language instructions. While these capabilities have led to widespread adoption, most LLMs are developed by resource-rich organizations and are frequently kept from the public. As a step towards democratizing this powerful technology, we present BLOOM, a 176B-parameter open-access language model designed and built thanks to a collaboration of hundreds of researchers. BLOOM is a decoder-only Transformer language model that was trained on the ROOTS corpus, a dataset comprising hundreds of sources in 46 natural and 13 programming languages (59 in total). We find that BLOOM achieves competitive performance on a wide variety of benchmarks, with stronger results after undergoing multitask prompted finetuning. To facilitate future research and applications using LLMs, we publicly release our models and code under the Responsible AI License.
translated by 谷歌翻译
Developing safe and useful general-purpose AI systems will require us to make progress on scalable oversight: the problem of supervising systems that potentially outperform us on most skills relevant to the task at hand. Empirical work on this problem is not straightforward, since we do not yet have systems that broadly exceed our abilities. This paper discusses one of the major ways we think about this problem, with a focus on how to turn it into one that can be productively studied empirically. We first present an experimental design centered on choosing tasks for which human specialists succeed but unaided humans and current general AI systems fail. We then present a proof-of-concept experiment following meant to demonstrate a key feature of this experimental design and show its viability with two question-answering tasks: MMLU and time-limited QuALITY. On these tasks, we find that human participants who interact with an unreliable large-language-model dialog assistant through chat -- a trivial baseline strategy for scalable oversight -- substantially outperform both the model alone and their own unaided performance. These results are an encouraging sign that scalable oversight will be tractable to study with present models and bolster recent findings that large language models can productively assist humans with difficult tasks.
translated by 谷歌翻译
边缘计算是一项有前途的技术,可以在需要瞬时数据处理的技术领域提供新功能。机器和深度学习等领域的研究人员对其应用程序进行了广泛的边缘和云计算,这主要是由于他们提供的大量计算和存储资源。目前,机器人技术也正在寻求利用这些功能,并且随着5G网络的开发,可以克服该领域的一些现有限制。在这种情况下,重要的是要知道如何利用新兴的边缘体系结构,当今存在哪些类型的边缘体系结构和平台,以及哪些可以并且应该基于每个机器人应用程序使用。一般而言,边缘平台可以以不同的方式实现和使用,尤其是因为有几个提供商提供或多或少提供的一组服务以及一些基本差异。因此,本研究针对那些从事下一代机器人系统开发的人解决了这些讨论,并将有助于理解每个边缘计算体系结构的优势和缺点,以便明智地选择适合每个应用程序的功能。
translated by 谷歌翻译
我们研究语言模型是否可以评估自己主张的有效性,并预测他们能够正确回答的问题。我们首先表明,当以正确的格式提供时,较大的模型在多样化的多项选择和True/False问题上进行了很好的校准。因此,我们可以通过要求模型首先提出答案,然后评估其答案正确的概率“ p(true)”来对开放式采样任务进行自我评估。我们发现在各种任务中,P(true)的表现,校准和缩放令人鼓舞。当我们允许模型考虑自己的许多样本之前,在预测一种特定可能性的有效性之前,自我评估的性能进一步改善。接下来,我们研究是否可以培训模型来预测“ P(ik)”,即“我知道”问题的概率,而无需参考任何特定提出的答案。模型在预测P(IK)方面表现良好,并且在跨任务中部分概括,尽管它们在新任务上的P(IK)校准方面遇到了困难。预测的p(IK)概率在存在相关的原始材料的情况下以及对数学单词问题解决方案的提示也适当增加。我们希望这些观察结果为培训更诚实的模型提供了基础,并研究了诚实对模型模仿人类写作以外的其他目标培训的案例的普遍性。
translated by 谷歌翻译
在最大的状态熵探索框架中,代理商与无奖励环境进行交互,以学习最大程度地提高其正在引起的预期国有访问的熵的政策。 Hazan等。 (2019年)指出,马尔可夫随机策略类别足以满足最大状态熵目标,而在这种情况下,利用非马克维亚性通常被认为是毫无意义的。在本文中,我们认为非马克维亚性是有限样本制度中最大状态熵探索至关重要的。尤其是,我们重新阐明了目标在一次试验中针对诱发的国有访问的预期熵的目标。然后,我们表明,非马克维亚确定性政策的类别足以满足引入的目标,而马尔可夫政策总体上遭受了非零的遗憾。但是,我们证明找到最佳的非马克维亚政策的问题是NP-HARD。尽管结果有负面的结果,但我们讨论了以一种可行的方式解决该问题的途径,以及非马克维亚探索如何使未来工作中在线增强学习的样本效率受益。
translated by 谷歌翻译
期望最大化(EM)算法是潜在变量模型推断的默认算法。与任何其他机器学习领域一样,潜在变量模型到非常大的数据集的应用使得使用高级并行和分布式架构。本文介绍了FEDEM,它是EM算法到联合学习背景的第一个扩展。 FEDEM是一种新的通信高效方法,其处理本地设备的部分参与,并且对数据集的异构分布具有稳健。为了缓解通信瓶颈,FedEM压缩适当定义的完整数据足够的统计数据。我们还开发并分析了FEDEM的延伸,以进一步纳入方差减少方案。在所有情况下,我们都会导出有限时间的复杂性范围,以便平滑非凸起问题。提出了数值结果以支持我们的理论发现,以及对生物多样性监测的联合缺失值估算的应用。
translated by 谷歌翻译